Goto

Collaborating Authors

 Center


TEMPO: Global Temporal Building Density and Height Estimation from Satellite Imagery

Glazer, Tammy, Hacheme, Gilles Q., Zaytar, Akram, Marotti, Luana, Michaels, Amy, Tadesse, Girmaw Abebe, White, Kevin, Dodhia, Rahul, Zolli, Andrew, Becker-Reshef, Inbal, Ferres, Juan M. Lavista, Robinson, Caleb

arXiv.org Artificial Intelligence

We present TEMPO, a global, temporally resolved dataset of building density and height derived from high-resolution satellite imagery using deep learning models. We pair building footprint and height data from existing datasets with quarterly PlanetScope basemap satellite images to train a multi-task deep learning model that predicts building density and building height at a 37.6-meter per pixel resolution. We apply this model to global PlanetScope basemaps from Q1 2018 through Q2 2025 to create global, temporal maps of building density and height. We validate these maps by comparing against existing building footprint datasets. Our estimates achieve an F1 score between 85% and 88% on different hand-labeled subsets, and are temporally stable, with a 0.96 five-year trend-consistency score. TEMPO captures quarterly changes in built settlements at a fraction of the computational cost of comparable approaches, unlocking large-scale monitoring of development patterns and climate impacts essential for global resilience and adaptation efforts.


Jennifer Lawrence Goes Dark

The New Yorker

She has been cast in maternal roles since her teens. Now, playing a mother for the first time since becoming one, she has chosen the part of a woman pushed past the edge of sanity. In "Die My Love," Lawrence, as Grace, vibrates with boredom and fury. The novel "Die, My Love," by the Argentinean writer Ariana Harwicz, is narrated by a wife and new mother who is living in rural France and seems to be losing her mind. Motherhood has inserted an immersion blender into her psyche: lust, repulsion, pleasure, and doom swirl into a single mess. She calls herself a "sodomising rodent" with "bullet-wounds for eyes," and thinks, "When I masturbate I desecrate crypts, and when I rock my baby I say amen, and when I smile I unplug an iron lung." One night, standing in the cold, staring at her family through a sliding door, she thinks, "I'll stop trying to draw blood from a stone. I'll contain my madness, I'll use the bathroom. I'll put my baby to sleep, jerk off my man and postpone my rebellion in favor of a better life." Martin Scorsese saw a brief review of the novel in the some years ago and decided to pick up a copy. He found it to be a "powerful mosaic of the mind," he told me recently. Scorsese is a member of a book club of sorts, with a few other filmmakers, who read with an eye toward adaptation. For "Die, My Love," he imagined casting Jennifer Lawrence in the lead. He'd been amazed by her performance in Darren Aronofsky's bewildering 2017 fantasia, "Mother!" In that surreal film--it's like an allegory set inside an oil painting--Lawrence plays a woman living with her poet husband in an old farmhouse, which is gradually, then apocalyptically, invaded by strangers. "She really is feeling everything that's happening, in what appears to be a dream of some kind," Scorsese said. He and Lawrence had discussed adaptations before. They considered "The Awakening," Kate Chopin's 1899 novel of female liberation, which ends with the protagonist, Edna Pontellier, walking into the sea. "Die, My Love" was like "The Awakening" if it began with Edna already underwater.


Finding My Voice: Generative Reconstruction of Disordered Speech for Automated Clinical Evaluation

Rosero, Karen, Yeo, Eunjung, Mortensen, David R., Slot, Cortney Van't, Hallac, Rami R., Busso, Carlos

arXiv.org Artificial Intelligence

ABSTRACT We present ChiReSSD, a speech reconstruction framework that preserves children speaker's identity while suppressing mispronunciations. Unlike prior approaches trained on healthy adult speech, ChiReSSD adapts to the voices of children with speech sound disorders (SSD), with particular emphasis on pitch and prosody. We evaluate our method on the ST AR dataset and report substantial improvements in lexical accuracy and speaker identity preservation. Furthermore, we automatically predict the phonetic content in the original and reconstructed pairs, where the proportion of corrected consonants is comparable to the percentage of correct consonants (PCC), a clinical speech assessment metric. Our experiments show Pearson correlation of ρ = 0.63 between automatic and human expert annotations, highlighting the potential to reduce the manual transcription burden. In addition, experiments on the TORGO dataset demonstrate effective generalization for reconstructing adult dysarthric speech. Our results indicate that disentangled, style-based TTS reconstruction can provide identity-preserving speech across diverse clinical populations.


Survey of HPC in US Research Institutions

Shu, Peng, Chen, Junhao, Liu, Zhengliang, Zhao, Huaqin, Li, Xinliang, Liu, Tianming

arXiv.org Artificial Intelligence

The rapid growth of AI, data-intensive science, and digital twin technologies has driven an unprecedented demand for high-performance computing (HPC) across the research ecosystem. While national laboratories and industrial hyperscalers have invested heavily in exascale and GPU-centric architectures, university-operated HPC systems remain comparatively under-resourced. This survey presents a comprehensive assessment of the HPC landscape across U.S. universities, benchmarking their capabilities against Department of Energy (DOE) leadership-class systems and industrial AI infrastructures. We examine over 50 premier research institutions, analyzing compute capacity, architectural design, governance models, and energy efficiency. Our findings reveal that university clusters, though vital for academic research, exhibit significantly lower growth trajectories (CAGR $\approx$ 18%) than their national ($\approx$ 43%) and industrial ($\approx$ 78%) counterparts. The increasing skew toward GPU-dense AI workloads has widened the capability gap, highlighting the need for federated computing, idle-GPU harvesting, and cost-sharing models. We also identify emerging paradigms, such as decentralized reinforcement learning, as promising opportunities for democratizing AI training within campus environments. Ultimately, this work provides actionable insights for academic leaders, funding agencies, and technology partners to ensure more equitable and sustainable HPC access in support of national research priorities.


Curate, Connect, Inquire: A System for Findable Accessible Interoperable and Reusable (FAIR) Human-Robot Centered Datasets

Zhou, Xingru, Modak, Sadanand, Chan, Yao-Cheng, Deng, Zhiyun, Sentis, Luis, Esteva, Maria

arXiv.org Artificial Intelligence

--The rapid growth of AI in robotics has amplified the need for high-quality, reusable datasets, particularly in human-robot interaction (HRI) and AI-embedded robotics. While more robotics datasets are being created, the landscape of open data in the field is uneven. This is due to a lack of curation standards and consistent publication practices, which makes it difficult to discover, access, and reuse robotics data. T o address these challenges, this paper presents a curation and access system with two main contributions: (1) a structured methodology to curate, publish, and integrate F AIR (Findable, Accessible, Interoperable, Reusable) human-centered robotics datasets; and (2) a ChatGPT -powered conversational interface trained with the curated datasets metadata and documentation to enable exploration, comparison robotics datasets and data retrieval using natural language. Developed based on practical experience curating datasets from robotics labs within T exas Robotics at the University of T exas at Austin, the system demonstrates the value of standardized curation and persistent publication of robotics data. The system's evaluation suggests that access and understandability of human-robotics data are significantly improved. This work directly aligns with the goals of the HCRL @ ICRA 2025 workshop and represents a step towards more human-centered access to data for embodied AI. I. INTRODUCTION The rise of AI-embedded robotics has made the need for high-quality datasets for varied training applications critical. In response, researchers are increasingly creating datasets specifically for usage in AI applications. Derived from complex and often interdisciplinary studies using mixed research methods, these often large and multimodal datasets reflect both the robots' and the humans' perspectives; some gathered in the context of carefully designed experiments and others during observations in the physical world.


Rapid morphology characterization of two-dimensional TMDs and lateral heterostructures based on deep learning

He, Junqi, Zhang, Yujie, Wang, Jialu, Wang, Tao, Zhang, Pan, Cai, Chengjie, Yang, Jinxing, Lin, Xiao, Yang, Xiaohui

arXiv.org Artificial Intelligence

Leveraging advancements in artificial intelligence, we introduce a deep learning-based method for efficiently characterizing heterostructures and 2D materials, specifically MoS 2-MoSe 2 lateral heterostructures and MoS 2 flakes with varying shapes and thicknesses. By utilizing YOLO models, we achieve an accuracy rate of over 94.67% in identifying these materials. Additionally, we explore the application of transfer learning across different materials, which further enhances model performance. This model exhibits robust generalization and anti-interference ability, ensuring reliable results in diverse scenarios. To facilitate practical use, we have developed an application that enables real-time analysis directly from optical microscope images, making the process significantly faster and more cost-effective than traditional methods. This deep learning-driven approach represents a promising tool for the rapid and accurate characterization of 2D materials, opening new avenues for research and development in material science. Keywords 2D material, TMDs, lateral heterostructure, deep learning, instance segmentation, morphology characterization Introduction Two-dimensional (2D) materials have attracted significant attention due to their excellent mechanical, electrical, thermal, and optical properties, making them ideal candidates for next-generation technologies.


Position: Stop Acting Like Language Model Agents Are Normal Agents

Perrier, Elija, Bennett, Michael Timothy

arXiv.org Artificial Intelligence

Language Model Agents (LMAs) are increasingly treated as capable of autonomously navigating interactions with humans and tools. Their design and deployment tends to presume they are normal agents capable of sustaining coherent goals, adapting across contexts and acting with a measure of intentionality. These assumptions are critical to prospective use cases in industrial, social and governmental settings. But LMAs are not normal agents. They inherit the structural problems of the large language models (LLMs) around which they are built: hallucinations, jailbreaking, misalignment and unpredictability. In this Position paper we argue LMAs should not be treated as normal agents, because doing so leads to problems that undermine their utility and trustworthiness. We enumerate pathologies of agency intrinsic to LMAs. Despite scaffolding such as external memory and tools, they remain ontologically stateless, stochastic, semantically sensitive, and linguistically intermediated. These pathologies destabilise the ontological properties of LMAs including identifiability, continuity, persistence and and consistency, problematising their claim to agency. In response, we argue LMA ontological properties should be measured before, during and after deployment so that the negative effects of pathologies can be mitigated.


LLM-TA: An LLM-Enhanced Thematic Analysis Pipeline for Transcripts from Parents of Children with Congenital Heart Disease

Raza, Muhammad Zain, Xu, Jiawei, Lim, Terence, Boddy, Lily, Mery, Carlos M., Well, Andrew, Ding, Ying

arXiv.org Artificial Intelligence

Thematic Analysis (TA) is a fundamental method in healthcare research for analyzing transcript data, but it is resource-intensive and difficult to scale for large, complex datasets. This study investigates the potential of large language models (LLMs) to augment the inductive TA process in high-stakes healthcare settings. Focusing on interview transcripts from parents of children with Anomalous Aortic Origin of a Coronary Artery (AAOCA), a rare congenital heart disease, we propose an LLM-Enhanced Thematic Analysis (LLM-TA) pipeline. Our pipeline integrates an affordable state-of-the-art LLM (GPT-4o mini), LangChain, and prompt engineering with chunking techniques to analyze nine detailed transcripts following the inductive TA framework. We evaluate the LLM-generated themes against human-generated results using thematic similarity metrics, LLM-assisted assessments, and expert reviews. Results demonstrate that our pipeline outperforms existing LLM-assisted TA methods significantly. While the pipeline alone has not yet reached human-level quality in inductive TA, it shows great potential to improve scalability, efficiency, and accuracy while reducing analyst workload when working collaboratively with domain experts. We provide practical recommendations for incorporating LLMs into high-stakes TA workflows and emphasize the importance of close collaboration with domain experts to address challenges related to real-world applicability and dataset complexity. https://github.com/jiaweixu98/LLM-TA


Mixed Effects Deep Learning for the interpretable analysis of single cell RNA sequencing data by quantifying and visualizing batch effects

Andrade, Aixa X., Nguyen, Son, Montillo, Albert

arXiv.org Artificial Intelligence

Single-cell RNA sequencing (scRNA-seq) data are often confounded by technical or biological batch effects. Existing deep learning models mitigate these effects but often discard batch-specific information, potentially losing valuable biological insights. We propose a Mixed Effects Deep Learning (MEDL) autoencoder framework that separately models batch-invariant (fixed effects) and batch-specific (random effects) components. By decoupling batch-invariant biological states from batch variations, our framework integrates both into predictive models. Our approach also generates 2D visualizations of how the same cell appears across batches, enhancing interpretability. Retaining both fixed and random effect latent spaces improves classification accuracy. We applied our framework to three datasets spanning the cardiovascular system (Healthy Heart), Autism Spectrum Disorder (ASD), and Acute Myeloid Leukemia (AML). With 147 batches in the Healthy Heart dataset, far exceeding typical numbers, we tested our framework's ability to handle many batches. In the ASD dataset, our approach captured donor heterogeneity between autistic and healthy individuals. In the AML dataset, it distinguished donor heterogeneity despite missing cell types and diseased donors exhibiting both healthy and malignant cells. These results highlight our framework's ability to characterize fixed and random effects, enhance batch effect visualization, and improve prediction accuracy across diverse datasets.


Machine Learning for Asymptomatic Ratoon Stunting Disease Detection With Freely Available Satellite Based Multispectral Imaging

Waters, Ethan Kane, Chen, Carla Chia-ming, Azghadi, Mostafa Rahimi

arXiv.org Artificial Intelligence

Disease detection in sugarcane, particularly the identification of asymptomatic infectious diseases such as Ratoon Stunting Disease (RSD), is critical for effective crop management. This study employed various machine learning techniques to detect the presence of RSD in different sugarcane varieties, using vegetation indices derived from freely available satellite-based spectral data. Our results show that the Support Vector Machine with a Radial Basis Function Kernel (SVM-RBF) was the most effective algorithm, achieving classification accuracy between 85.64% and 96.55%, depending on the variety. Gradient Boosting and Random Forest also demonstrated high performance achieving accuracy between 83.33% to 96.55%, while Logistic Regression and Quadratic Discriminant Analysis showed variable results across different varieties. The inclusion of sugarcane variety and vegetation indices was important in the detection of RSD. This agreed with what was identified in the current literature. Our study highlights the potential of satellite-based remote sensing as a cost-effective and efficient method for large-scale sugarcane disease detection alternative to traditional manual laboratory testing methods.